STDPboost: A Self-Training Method Based on Density Peaks and Improved Adaboost for Semi-Supervised Classification
نویسندگان
چکیده
The self-training methods have been praised by extensive research in semi-supervised classification. Mislabeling is the main challenge methods. Multiple variations of are recently proposed against mislabeling from following one two aspects: a) using heuristic rules to find high-confidence unlabeled samples that can easily be predicted correctly each iteration; b) enhancing prediction performance employing ensemble classifiers composed multiple weak classifiers. Yet, they still suffer issues: a); most strategies for finding heavily rely on parameters; almost all employed originally designed supervised and may not suitable classification due limited number unrepresented distribution initial labeled data; c) few overcome above aspects at same time. To advance state art, a new method based density peaks clustering improved Adaboost presented named as STDPboost. In iterative self-taught process, clustering-based strategy classifier AdaboostSEMI more predict samples, which overcomes mentioned shortcomings existing Intensive experiments benchmark data sets proven STDPboost outperforms 7 state-of-the-art average accuracy KNN CART with percentages 10% 50% further alleviating mislabeling.
منابع مشابه
Semi-supervised Affinity Propagation Based on Density Peaks
Original scientific paper In view of the unsatisfying clustering effect of affinity propagation (AP) clustering algorithm when dealing with data sets of complex structures, a semi-supervised affinity propagation clustering algorithm based on density peaks (SAP-DP) was proposed in this paper. The algorithm uses a new algorithm of density peaks (DP) which has the advantage of the manifold cluster...
متن کاملA Self-training Method for Semi-supervised Gans
Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...
متن کاملA Self-Training Method for Semi-Supervised GANs
Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...
متن کاملA Self-training Method for Semi-supervised Gans
Since the creation of Generative Adversarial Networks (GANs), much work has been done to improve their training stability, their generated image quality, their range of application but nearly none of them explored their self-training potential. Self-training has been used before the advent of deep learning in order to allow training on limited labelled training data and has shown impressive res...
متن کاملSemi-supervised Self-training for Sentence Subjectivity Classification
Recent natural language processing (NLP) research shows that identifying and extracting subjective information from texts can benefit many NLP applications. In this paper, we address a semi-supervised learning approach, self-training, for sentence subjectivity classification. In self-training, the confidence degree that depends on the ranking of class membership probabilities is commonly used a...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Access
سال: 2023
ISSN: ['2169-3536']
DOI: https://doi.org/10.1109/access.2023.3294982